home
***
CD-ROM
|
disk
|
FTP
|
other
***
search
/
Languguage OS 2
/
Languguage OS II Version 10-94 (Knowledge Media)(1994).ISO
/
gnu
/
glibc108.zip
/
glibc108
/
sysdeps
/
i386
/
add_n.S
< prev
next >
Wrap
Text File
|
1994-05-16
|
2KB
|
94 lines
# i80386 __mpn_add_n -- Add two limb vectors of the same length > 0 and store
# sum in a third limb vector.
# Copyright (C) 1992, 1994 Free Software Foundation, Inc.
# This file is part of the GNU MP Library.
# The GNU MP Library is free software; you can redistribute it and/or modify
# it under the terms of the GNU Library General Public License as published by
# the Free Software Foundation; either version 2 of the License, or (at your
# option) any later version.
# The GNU MP Library is distributed in the hope that it will be useful, but
# WITHOUT ANY WARRANTY; without even the implied warranty of MERCHANTABILITY
# or FITNESS FOR A PARTICULAR PURPOSE. See the GNU Library General Public
# License for more details.
# You should have received a copy of the GNU Library General Public License
# along with the GNU MP Library; see the file COPYING.LIB. If not, write to
# the Free Software Foundation, Inc., 675 Mass Ave, Cambridge, MA 02139, USA.
# INPUT PARAMETERS
# res_ptr (sp + 4)
# s1_ptr (sp + 8)
# s2_ptr (sp + 12)
# size (sp + 16)
#include "asm.h"
#include "sysdep.h"
.text
.align 2
.globl C_SYMBOL_NAME(__mpn_add_n)
C_SYMBOL_NAME(__mpn_add_n):
pushl %edi
pushl %esi
movl 12(%esp),%edi # res_ptr
movl 16(%esp),%esi # s1_ptr
movl 20(%esp),%edx # s2_ptr
movl 24(%esp),%ecx # size
movl %ecx,%eax
shrl $3,%ecx # compute loop count for unrolled loop
negl %eax
andl $7,%eax # get index where to start loop
jz Loop # necessary special case for 0
incl %ecx # adjust loop count
shll $2,%eax # adjustment for pointers...
subl %eax,%edi # ... since they are offset ...
subl %eax,%esi # ... by a constant when we enter ...
subl %eax,%edx # ... the loop
shrl $2,%eax # restore previous value
leal (Loop - 3)(%eax,%eax,8),%eax # calc start addr in loop
jmp *%eax # jump into loop
.align 2
Loop: movl (%esi),%eax
adcl (%edx),%eax
movl %eax,(%edi)
movl 4(%esi),%eax
adcl 4(%edx),%eax
movl %eax,4(%edi)
movl 8(%esi),%eax
adcl 8(%edx),%eax
movl %eax,8(%edi)
movl 12(%esi),%eax
adcl 12(%edx),%eax
movl %eax,12(%edi)
movl 16(%esi),%eax
adcl 16(%edx),%eax
movl %eax,16(%edi)
movl 20(%esi),%eax
adcl 20(%edx),%eax
movl %eax,20(%edi)
movl 24(%esi),%eax
adcl 24(%edx),%eax
movl %eax,24(%edi)
movl 28(%esi),%eax
adcl 28(%edx),%eax
movl %eax,28(%edi)
leal 32(%edi),%edi
leal 32(%esi),%esi
leal 32(%edx),%edx
decl %ecx
jnz Loop
sbbl %eax,%eax
negl %eax
popl %esi
popl %edi
ret